Discrete-Time Markovian Decision Processes with Incomplete State Observation
نویسندگان
چکیده
منابع مشابه
Learning Without State-Estimation in Partially Observable Markovian Decision Processes
Reinforcement learning RL algorithms pro vide a sound theoretical basis for building learning control architectures for embedded agents Unfortunately all of the theory and much of the practice see Barto et al for an exception of RL is limited to Marko vian decision processes MDPs Many real world decision tasks however are inherently non Markovian i e the state of the environ ment is only incomp...
متن کاملStability of Markovian Processes I: Criteria for Discrete-time Chains
In this paper we connect various topological and probabilistic forms of stability for discrete-time Markov chains. These include tightness on the one hand and Harris recurrence and ergodicity on the other. We show that these concepts of stability are largely equivalent for a major class of chains (chains with continuous components), or if the state space has a sufficiently rich class of appropr...
متن کاملA discrete-state discrete-time model using indirect observation.
This research was motivated by a desire to model the progression of a chronic disease through various disease stages when data are not available to directly estimate all the transition parameters in the model. This is a common occurrence when time and expense make it unfeasible to follow a single cohort to estimate all the transition parameters. One difficulty of developing a model of chronic d...
متن کاملEigenvalue Assignment Of Discrete-Time Linear Systems With State And Input Time-Delays
Time-delays are important components of many dynamical systems that describe coupling or interconnection between dynamics, propagation or transport phenomena, and heredity and competition in population dynamics. The stabilization with time delay in observation or control represents difficult mathematical challenges in the control of distributed parameter systems. It is well-known that the stabi...
متن کاملMarkov decision processes with observation costs
A partially observable Markov decision process (POMDP) is a generalization of a Markov decision process in which observation of the process state can be imperfect and/or costly. Although it provides an elegant model for control and planning problems that include information-gathering actions, the best current algorithms for POMDPs are computationally infeasible for all but small problems. One a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: The Annals of Mathematical Statistics
سال: 1970
ISSN: 0003-4851
DOI: 10.1214/aoms/1177697190